In computing, memory addresses are index-like numbers used at various levels by software and hardware to access the computer's primary storage memory.
Contents |
There are several types of memory addresses. In other words, a computer, and even one program may have several different memory address spaces.
A digital computer's memory, more specifically main memory, consists of many memory locations, each having a physical address, a code, which the CPU can use to access it. Generally only system software, i.e. the BIOS, operating systems, and some specialized utility programs (e.g., memory testers), address physical memory using machine code operands or processor registers, instructing the CPU to direct a hardware device, called the memory controller, to use the memory bus or system bus, or separate control, address and data busses, to execute the program's commands. The memory controllers' bus consists of a number of parallel lines, each represented by a binary digit (bit). The width of the bus, and thus the number of addressable storage units, and the number of bits in each unit, varies among computers.
A computer program use memory addresses to execute machine code, store and retrieve data. Most application programs do not have indeed a knowledge about physical addresses. Rather, they address logical addresses, or virtual addresses, using computer's and operating system memory mapping.
Most modern computers are byte-addressable, with each address identifying a single eight bit byte or character of storage; data too large to be stored in a single byte may reside in multiple bytes occupying a sequence of consecutive addresses. Other computers are word-addressable, with addressable storage units larger than a byte. For example, the Data General Nova minicomputer, and the Texas Instruments TMS9900 and National Semiconductor IMP-16 microcomputers used 16 bit words, and there were many 36-bit mainframe computers which used 18-bit word addressing. The efficiency of addressing of memory depends on the bit size of the bus used for addresses—the more bits used, the more addresses are available to the computer. For example, a byte-addressable machine with a 20-bit address bus (e.g. Intel 8086) can address 220 (1,048,576) memory locations, or one megabyte of memory, while a 32-bit bus (e.g. Intel 80386) addresses 232 (4,294,967,296) locations, or a 4 GiB address space. In contrast, a 36-bit word-addressable machine with an 18-bit bus addresses only 218 (262,144) locations, but 1.125 megabytes—slightly more than the 8086.
Some older computers (decimal computers), were decimal digit-addressable. For example, each address in the IBM 1620's magnetic-core memory identified a single six bit binary-coded decimal digit, consisting of a parity bit, flag bit and four numerical bits. The 1620 used 5-digit decimal addresses, so in theory the highest possible address was 99,999. In practice, the CPU supported 20,000 memory locations, and up to two optional external memory units could be added, each supporting 20,000 addresses, for a total of 60,000 (00000–59999).
Each memory location in a stored-program computer holds a binary number of some sort. Its interpretation, as data of some data type or as an instruction, and use are determined by the instructions which retrieve and manipulate it.
Some early programmers combined instructions and data in words as a way to save memory, when it was expensive: The Manchester Mark 1 had space in its 40-bit words to store little bits of data— its processor ignored a small section in the middle of a word— and that was often exploited as extra data storage. Self-replicating programs such as viruses treat themselves sometimes as data and sometimes as instructions. Self-modifying code is generally deprecated nowadays, as it makes testing and maintenance disproportionally difficult to the saving of a few bytes, and can also give incorrect results because of the compiler or processor's assumptions about the machine's state, but is still sometimes used deliberately, with great care.
A word size is characteristic to a given computer architecture. It denotes the number of bits that a CPU can process at one time. Historically it has been sized in multiples of four and eight bits (nibbles and bytes, respectively), so sizes of 4, 8, 12, 16, 24, 32, 48, 64, and larger came into vogue with technological advances.
Very often, when referring to the word size of a modern computer, one is also describing the size of address space on that computer. For instance, a computer said to be "32-bit" also usually allows 32-bit memory addresses; a byte-addressable 32-bit computer can address bytes of memory, or 4 gibibytes (GiB). This seems logical and useful, as it allows one memory address to be efficiently stored in one word.
But this does not always hold. Computers often have memory addresses larger or smaller than their word size. For instance, almost all 8-bit processors, such as 6502, supported 16-bit addresses— if not they would have been limited to a mere 256 byte memory. The 16-bit Intel 8088 had only an 8-bit external memory bus on early IBM PCs, and the 16-bit Intel 8086 supported 20-bit addressing, allowing it to access 1 MiB rather than 64 KiBs of memory. Popular Intel Pentium processors since introduction of Physical Address Extensions (PAE) support 36-bit physical addresses, while generally having only a 32-bit word.
A modern byte-addressable 64-bit computer—with proper OS support can address bytes (or 16 exbibytes) which as of 2011[update] is considered practically unlimited.
A computer program can access an address given explicitly – in low-level programming this is usually called an absolute address, or sometimes a specific address, and is known as pointer data type in higher-level languages. But a program can also use relative address which specifies a location in relation to somewhere else (the base address). There are much more indirect addressing modes.
Mapping logical addresses to physical and virtual memory also adds several levels of indirection; see below.
Many programmers prefer to address memory such that there is no distinction between code space and data space, as well as from physical and virtual memory (see below) — in other words, numerically identical pointers refer to exactly the same byte of RAM.
However, many early computers did not support such a flat memory model — in particular, Harvard architecture machines force program storage to be completely separate from data storage. Many modern DSPs (such as the Motorola 56000) have 3 separate storage areas — program storage, coefficient storage, and data storage. Some commonly-used instructions fetch from all three areas simultaneously — fewer storage areas (even if there were the same total bytes of storage) would make those instructions run slower.
Early x86 computers used the segmented memory model addresses based on a combination of two numbers: a memory segment, and an offset within that segment. Some segments were implicitly treated as code segments, dedicated for instructions, stack segments, or normal data segments. Although the usages were different, the segments did not have different memory protections reflecting this. In the flat memory model all segments (segment registers) are generally set to zero, and only offsets are variable.
Virtual memory is now universally used on desktop machines. It maps physical memory to different addresses using page tables. By doing so the operating system can decide to allocate memory as it deems most efficient without causing the program to halt from a long garbage collection process. Some literature suggests that, on the whole, garbage collection is the most efficient of memory reclamation strategies , but it is not ideal if it cuts in to 'tidy up' memory just as the program is controlling the firing of a missile or a landing on the moon, because it is non-deterministic.
The physical blocks of memory (typically 4 kb chunks) are mapped to virtual addresses by a virtual memory program in the kernel, and supported by the processor hardware, though it was formerly done all in software. The purpose of virtual memory is to abstract memory allocation, allowing the physical space to be allocated as best for the hardware (that is, usually in good sized blocks), and to still be seen as contiguous from a program or compiler's perspective. Virtual memory is supported by some operating systems (for example Linux and Windows but not MS-DOS). One may think of virtual memory as a filter, or an alternate set of memory addresses that allow programs to read from memory as quickly as possible without requiring it to be at a particular place. Programs use these contiguous virtual addresses, rather than real, and often fragmented, physical addresses, to store instructions and data. When the program is actually executed, the virtual addresses are translated by the processor into real memory addresses. logical address is a synonym for virtual address.
Virtual memory also effectively allows the address space to be larger than – or rather, extend beyond – the amount of real (or physical) memory available; the computer can put rarely-accessed pages into secondary storage (similar, at a level, in behavior to CPU cache) and use the real memory (RAM) for new or active tasks. So the virtual address space might contain, say, twice as many addresses as main memory with the extra addresses mapped to hard disk space in the form of a swap file (used in Linux) or page file (the default on Windows XP and later MS operating systems). When required they are copied back (called swapping) into main memory as soon as they are needed. These movements are performed, typically, as a background process – and in that sense are transparent to programs. However, if due care is not taken, this can often lead to thrashing.